56 research outputs found
Regression-based Multi-View Facial Expression Recognition
We present a regression-based scheme for multi-view facial expression recognition based on 2-D geometric features. We address the problem by mapping facial points (e.g. mouth corners) from non-frontal to frontal view where further recognition of the expressions can be performed using a state-of-the-art facial expression recognition method. To learn the mapping functions we investigate four regression models: Linear Regression (LR), Support Vector Regression (SVR), Relevance Vector Regression (RVR) and Gaussian Process Regression (GPR). Our extensive experiments on the CMU Multi-PIE facial expression database show that the proposed scheme outperforms view-specific classifiers by utilizing considerably less training data
Personalized Automatic Estimation of Self-reported Pain Intensity from Facial Expressions
Pain is a personal, subjective experience that is commonly evaluated through
visual analog scales (VAS). While this is often convenient and useful,
automatic pain detection systems can reduce pain score acquisition efforts in
large-scale studies by estimating it directly from the participants' facial
expressions. In this paper, we propose a novel two-stage learning approach for
VAS estimation: first, our algorithm employs Recurrent Neural Networks (RNNs)
to automatically estimate Prkachin and Solomon Pain Intensity (PSPI) levels
from face images. The estimated scores are then fed into the personalized
Hidden Conditional Random Fields (HCRFs), used to estimate the VAS, provided by
each person. Personalization of the model is performed using a newly introduced
facial expressiveness score, unique for each person. To the best of our
knowledge, this is the first approach to automatically estimate VAS from face
images. We show the benefits of the proposed personalized over traditional
non-personalized approach on a benchmark dataset for pain analysis from face
images.Comment: Computer Vision and Pattern Recognition Conference, The 1st
International Workshop on Deep Affective Learning and Context Modelin
Unsupervised Domain Adaptation with Copula Models
We study the task of unsupervised domain adaptation, where no labeled data
from the target domain is provided during training time. To deal with the
potential discrepancy between the source and target distributions, both in
features and labels, we exploit a copula-based regression framework. The
benefits of this approach are two-fold: (a) it allows us to model a broader
range of conditional predictive densities beyond the common exponential family,
(b) we show how to leverage Sklar's theorem, the essence of the copula
formulation relating the joint density to the copula dependency functions, to
find effective feature mappings that mitigate the domain mismatch. By
transforming the data to a copula domain, we show on a number of benchmark
datasets (including human emotion estimation), and using different regression
models for prediction, that we can achieve a more robust and accurate
estimation of target labels, compared to recently proposed feature
transformation (adaptation) methods.Comment: IEEE International Workshop On Machine Learning for Signal Processing
201
RoboChain: A Secure Data-Sharing Framework for Human-Robot Interaction
Robots have potential to revolutionize the way we interact with the world
around us. One of their largest potentials is in the domain of mobile health
where they can be used to facilitate clinical interventions. However, to
accomplish this, robots need to have access to our private data in order to
learn from these data and improve their interaction capabilities. Furthermore,
to enhance this learning process, the knowledge sharing among multiple robot
units is the natural step forward. However, to date, there is no
well-established framework which allows for such data sharing while preserving
the privacy of the users (e.g., the hospital patients). To this end, we
introduce RoboChain - the first learning framework for secure, decentralized
and computationally efficient data and model sharing among multiple robot units
installed at multiple sites (e.g., hospitals). RoboChain builds upon and
combines the latest advances in open data access and blockchain technologies,
as well as machine learning. We illustrate this framework using the example of
a clinical intervention conducted in a private network of hospitals.
Specifically, we lay down the system architecture that allows multiple robot
units, conducting the interventions at different hospitals, to perform
efficient learning without compromising the data privacy.Comment: 7 pages, 6 figure
Multi-Conditional Latent Variable Model for Joint Facial Action Unit Detection
We propose a novel multi-conditional latent variable model for simultaneous facial feature fusion and detection of facial action units. In our approach we exploit the structure-discovery capabilities of generative models such as Gaussian processes, and the discriminative power of classifiers such as logistic function. This leads to superior performance compared to existing classifiers for the target task that exploit either the discriminative or generative property, but not both. The model learning is performed via an efficient, newly proposed Bayesian learning strategy based on Monte Carlo sampling. Consequently, the learned model is robust to data overfitting, regardless of the number of both input features and jointly estimated facial action units. Extensive qualitative and quantitative experimental evaluations are performed on three publicly available datasets (CK+, Shoulder-pain and DISFA). We show that the proposed model outperforms the state-of-the-art methods for the target task on (i) feature fusion, and (ii) multiple facial action unit detection
- …